Multitask Fine-Tuning for Passage Re-Ranking Using BM25 and Pseudo Relevance Feedback

نویسندگان

چکیده

Passage re-ranking is a machine learning task that estimates relevance scores between given query and candidate passages. Keyword features based on the lexical similarities queries passages have been traditionally used for passage models. However, such approaches limitation; it difficult to find semantic contextual beyond word-matching information. Recently, several studies neural pre-trained language models as BERT overcome limitations of traditional keyword-based they show significant performance improvements. Such ranking advantage finding documents better than methods. these deep learning-based require large amounts data training. training usually manually labeled with high cost, how utilize efficiently an important issue. This paper proposes fine-tuning method efficient model. The proposed model utilizes augmentation by simultaneously MLM tasks during stages. For task, different parts are masked at each epoch. Even if only one pair given, exposed diverse cases dynamically from one. Also, probability distribution term importance trained model.We calculate weight two novel methods using BM25 pseudo feedback. Terms sampled according weight. learns representation executing task. A feedback applied calculating importance. It enables form feedbacks initial search stage. MS MARCO leaderboard Our achieves state-of-the-art MRR@10 score in except ensemble-based method. In addition, our demonstrates three evaluation metrics: MRR@10, Mean Rank, Hit@(5,10,20,50).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Pseudo-Relevance Feedback and Title Re-Ranking for Chinese Information Retrieval

In our formal runs, we have experimented with the retrieval based on character-based indexing and hybrid term indexing because these are more distinct types of indexing for better pooling. We confirmed that character-based indexing did not produce relatively good retrieval effectiveness. We have also experimented with three new pseudo-relevance feedback (PRF) methods. These new methods were abl...

متن کامل

On Improving Pseudo-Relevance Feedback Using Pseudo-Irrelevant Documents

Pseudo-Relevance Feedback (PRF) assumes that the topranking n documents of the initial retrieval are relevant and extracts expansion terms from them. In this work, we introduce the notion of pseudo-irrelevant documents, i.e. high-scoring documents outside of top n that are highly unlikely to be relevant. We show how pseudo-irrelevant documents can be used to extract better expansion terms from ...

متن کامل

Pseudo Relevance Feedback Using Named Entities for Question Answering

Relevance feedback has already proven its usefulness in probabilistic information retrieval (IR). In this research we explore whether a pseudo relevance feedback technique on IR can improve the Question Answering task (QA). The basis of our exploration is the use of relevant named entities from the top retrieved documents as clues of relevance. We discuss two interesting findings from these exp...

متن کامل

Extract-biased pseudo-relevance feedback

Successfully retrieving a web document is a twofold problem: having an adequate query that can usefully and properly help filtering relevant documents from huge collections, and presenting the user those that may indeed fulfill his/her needs. In this paper, we focus on the first issue – the problem of having a misleading user query. The aim of the work is to refine a query by using extracts ins...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2022

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2022.3176894